1 |
Universals of Linguistic Idiosyncrasy in Multilingual Computational Linguistics ; Universals of Linguistic Idiosyncrasy in Multilingual Computational Linguistics: Dagstuhl Seminar 21351
|
|
|
|
In: Universals of Linguistic Idiosyncrasy in Multilingual Computational Linguistics ; https://hal.archives-ouvertes.fr/hal-03507948 ; Universals of Linguistic Idiosyncrasy in Multilingual Computational Linguistics, Aug 2021, pp.89--138, 2021, 2192-5283. ⟨10.4230/DagRep.11.7.89⟩ ; https://gitlab.com/unlid/dagstuhl-seminar/-/wikis/home (2021)
|
|
BASE
|
|
Show details
|
|
2 |
Universals of Linguistic Idiosyncrasy in Multilingual Computational Linguistics (Dagstuhl Seminar 21351)
|
|
|
|
BASE
|
|
Show details
|
|
7 |
Syntactic Nuclei in Dependency Parsing -- A Multilingual Exploration ...
|
|
|
|
BASE
|
|
Show details
|
|
8 |
Revisiting Negation in Neural Machine Translation ...
|
|
|
|
Abstract:
Read paper: NA Abstract: In this paper, we evaluate the translation of negation both automatically and manually, in English—German (EN—DE) and English—Chinese (EN—ZH). We show that the ability of neural machine translation (NMT) models to translate negation has improved with deeper and more advanced networks, although the performance varies between language pairs and translation directions. The accuracy of manual evaluation in EN—DE, DE—EN, EN—ZH, and ZH—EN is 95.7%, 94.8%, 93.4%, and 91.7%, respectively. In addition, we show that under-translation is the most significant error type in NMT, which contrasts with the more diverse error profile previously observed for statistical machine translation. To better understand the root of the under-translation of negation, we study the model's information flow and training data. While our information flow analysis does not reveal any deficiencies that could be used to detect or fix the under-translation of negation, we find that negation is often rephrased during ...
|
|
Keyword:
Computational Linguistics; Condensed Matter Physics; Deep Learning; Electromagnetism; FOS Physical sciences; Information and Knowledge Engineering; Neural Network; Semantics
|
|
URL: https://underline.io/lecture/25803-revisiting-negation-in-neural-machine-translation https://dx.doi.org/10.48448/x27r-fa09
|
|
BASE
|
|
Hide details
|
|
9 |
Universals of Linguistic Idiosyncrasy in Multilingual Computational Linguistics (Dagstuhl Seminar 21351) ...
|
|
|
|
BASE
|
|
Show details
|
|
10 |
Attention Can Reflect Syntactic Structure (If You Let It) ...
|
|
|
|
BASE
|
|
Show details
|
|
11 |
What Should/Do/Can LSTMs Learn When Parsing Auxiliary Verb Constructions? ...
|
|
|
|
BASE
|
|
Show details
|
|
12 |
Schrödinger's Tree -- On Syntax and Neural Language Models ...
|
|
|
|
BASE
|
|
Show details
|
|
16 |
Køpsala: Transition-Based Graph Parsing via Efficient Training and Effective Encoding ...
|
|
|
|
BASE
|
|
Show details
|
|
17 |
Understanding Pure Character-Based Neural Machine Translation: The Case of Translating Finnish into English ...
|
|
|
|
BASE
|
|
Show details
|
|
18 |
Understanding Pure Character-Based Neural Machine Translation: The Case of Translating Finnish into English ...
|
|
|
|
BASE
|
|
Show details
|
|
19 |
Universal Dependencies v2: An Evergrowing Multilingual Treebank Collection ...
|
|
|
|
BASE
|
|
Show details
|
|
20 |
Do Neural Language Models Show Preferences for Syntactic Formalisms? ...
|
|
|
|
BASE
|
|
Show details
|
|
|
|